Communication and matrix computations on large message passing systems
نویسنده
چکیده
abstract This paper is concerned with the consequences for matrix computations of having a rather large number of general purpose processors, say ten or twenty thousand, connected in a network in such a way that a processor can communicate only with its immediate neighbors. Certain communication tasks associated with most matrix algorithms are deened and formulas developed for the time required to perform them under several communication regimes. The results are compared with the times for a nominal n 3 oating point operations. The results suggest that it is possible to use a large number of processors to solve matrix problems at a relatively ne granularity, provided ne grain communication is available.
منابع مشابه
Communication and Matrix Computations on Large Message Passing Systems Communication and Matrix Computations on Large Message Passing Systems
abstract This paper is concerned with the consequences for matrix computations of having a rather large number of general purpose processors, say ten or twenty thousand, connected in a network in such a way that a processor can communicate only with its immediate neighbors. Certain communication tasks associated with most matrix algorithms are deened and formulas developed for the time required...
متن کاملA Message-Passing Distributed Memory Parallel Algorithm for a Dual-Code Thin Layer, Parabolized Navier-Stokes Solver
In this study, the results of parallelization of a 3-D dual code (Thin Layer, Parabolized Navier-Stokes solver) for solving supersonic turbulent flow around body and wing-body combinations are presented. As a serial code, TLNS solver is very time consuming and takes a large part of memory due to the iterative and lengthy computations. Also for complicated geometries, an exceeding number of grid...
متن کاملParallel Reduction from Block Hessenberg to Hessenberg using MPI
In many scientific applications, eigenvalues of a matrix have to be computed. By first reducing a matrix from fully dense to Hessenberg form, eigenvalue computations with the QR algorithm become more efficient. Previous work on shared memory architectures has shown that the Hessenberg reduction is in some cases most efficient when performed in two stages: First reduce the matrix to block Hessen...
متن کاملPerformance Characterization of Matrix Multiplication on SGI Altix 3700
Matrix multiplication is widely used in a variety of applications and is often one of the core components of many scientific computations which includes graph theory, numerical methods, digital control and signal processing. Multiplication of large matrices require a lot of computation time as its complexity is O(n), where n is the dimension of the matrix. A serial algorithm to compute large ma...
متن کاملCompiling P–GRADE programs for the JGrid architecture∗
Grid systems are large, geographically distributed computational environments. JGrid is a Grid infrastructure, that provides a software environment for such systems. It is built on top of the Java-based Jini technology, which provides a common platform and a service-based lookup system. P–GRADE is an interactive, graphical software development tool, designed to build distributed parallel progra...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- Parallel Computing
دوره 16 شماره
صفحات -
تاریخ انتشار 1990